Goto

Collaborating Authors

 Pomerania Province


The toddler who survived a 54-degree body temperature

Popular Science

Humans aren't built for the cold, but have survived frigid temperatures in some amazing cases. Breakthroughs, discoveries, and DIY tips sent six days a week. Winter is not for the faint of heart. In New York City, skyscrapers turn Manhattan into a series of freezing wind tunnels. In Sapporo, Japan, the snowfall is almost 200 inches each winter. Even so, humans have developed plenty of clever ways to wait out the cold. But what would happen if instead of bundling up inside with a hot chocolate, you were left in the frigid cold--just how cold can humans get and recover?


VIVAT: Virtuous Improving VAE Training through Artifact Mitigation

Novitskiy, Lev, Vasilev, Viacheslav, Kovaleva, Maria, Arkhipkin, Vladimir, Dimitrov, Denis

arXiv.org Artificial Intelligence

Variational Autoencoders (VAEs) remain a cornerstone of generative computer vision, yet their training is often plagued by artifacts that degrade reconstruction and generation quality. This paper introduces VIVAT, a systematic approach to mitigating common artifacts in KL-VAE training without requiring radical architectural changes. We present a detailed taxonomy of five prevalent artifacts - color shift, grid patterns, blur, corner and droplet artifacts - and analyze their root causes. Through straightforward modifications, including adjustments to loss weights, padding strategies, and the integration of Spatially Conditional Normalization, we demonstrate significant improvements in VAE performance. Our method achieves state-of-the-art results in image reconstruction metrics (PSNR and SSIM) across multiple benchmarks and enhances text-to-image generation quality, as evidenced by superior CLIP scores. By preserving the simplicity of the KL-VAE framework while addressing its practical challenges, VIVAT offers actionable insights for researchers and practitioners aiming to optimize VAE training.


A Game-Theoretic Approach for Adversarial Information Fusion in Distributed Sensor Networks

Kallas, Kassem

arXiv.org Artificial Intelligence

Every day we share our personal information through digital systems which are constantly exposed to threats. For this reason, security-oriented disciplines of signal processing have received increasing attention in the last decades: multimedia forensics, digital watermarking, biometrics, network monitoring, steganography and steganalysis are just a few examples. Even though each of these fields has its own peculiarities, they all have to deal with a common problem: the presence of one or more adversaries aiming at making the system fail. Adversarial Signal Processing lays the basis of a general theory that takes into account the impact that the presence of an adversary has on the design of effective signal processing tools. By focusing on the application side of Adversarial Signal Processing, namely adversarial information fusion in distributed sensor networks, and adopting a game-theoretic approach, this thesis contributes to the above mission by addressing four issues. First, we address decision fusion in distributed sensor networks by developing a novel soft isolation defense scheme that protect the network from adversaries, specifically, Byzantines. Second, we develop an optimum decision fusion strategy in the presence of Byzantines. In the next step, we propose a technique to reduce the complexity of the optimum fusion by relying on a novel near-optimum message passing algorithm based on factor graphs. Finally, we introduce a defense mechanism to protect decentralized networks running consensus algorithm against data falsification attacks.


Compressing Sensor Data for Remote Assistance of Autonomous Vehicles using Deep Generative Models

Bogdoll, Daniel, Jestram, Johannes, Rauch, Jonas, Scheib, Christin, Wittig, Moritz, Zöllner, J. Marius

arXiv.org Artificial Intelligence

In the foreseeable future, autonomous vehicles will require human assistance in situations they can not resolve on their own. In such scenarios, remote assistance from a human can provide the required input for the vehicle to continue its operation. Typical sensors used in autonomous vehicles include camera and lidar sensors. Due to the massive volume of sensor data that must be sent in real-time, highly efficient data compression is elementary to prevent an overload of network infrastructure. Sensor data compression using deep generative neural networks has been shown to outperform traditional compression approaches for both image and lidar data, regarding compression rate as well as reconstruction quality. However, there is a lack of research about the performance of generative-neural-network-based compression algorithms for remote assistance. In order to gain insights into the feasibility of deep generative models for usage in remote assistance, we evaluate state-of-the-art algorithms regarding their applicability and identify potential weaknesses. Further, we implement an online pipeline for processing sensor data and demonstrate its performance for remote assistance using the CARLA simulator.




Supplementary Material for " Deep Learning with Label Differential Privacy " A Missing Proofs A.1 Proof of Lemma 1 Proof of Lemma 1

Neural Information Processing Systems

RRTop-k is " -DP as desired. The training set contains 60,000 examples and the test set contains 10,000. On MNIST, Fashion MNIST, and KMNIST, we train the models with mini-batch SGD with batch size 265 and momentum 0.9. On CIFAR-10, we use batch size 512 and momentum 0.9, and train for 200 epochs. The learning rate is scheduled according to the widely used piecewise constant with linear rampup scheme.


7c220a2091c26a7f5e9f1cfb099511e3-Supplemental.pdf

Neural Information Processing Systems

Appendix of "TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up" We also evaluate the effectiveness of stronger augmentation on high-resolution generative tasks (E.g. Table 1, 2, 3, 4. For the generator architectures, the "Block" represents the basic Transformer Block "Grid Block" denotes the Transformer Block where the standard self-attention is replaced by the propose For the discriminator architectures, we use "Layer Flatten" to represent the process of We compare the GPU memory cost between standard self-attention and grid self-attention. We evaluate the inference cost of these two architectures, without calculating the gradient. We include more high-resolution visual examples on Figure 3,4.



Model-Based Ranking of Source Languages for Zero-Shot Cross-Lingual Transfer

Ebrahimi, Abteen, Wiemerslage, Adam, von der Wense, Katharina

arXiv.org Artificial Intelligence

We present NN-Rank, an algorithm for ranking source languages for cross-lingual transfer, which leverages hidden representations from multilingual models and unlabeled target-language data. We experiment with two pretrained multilingual models and two tasks: part-of-speech tagging (POS) and named entity recognition (NER). We consider 51 source languages and evaluate on 56 and 72 target languages for POS and NER, respectively. When using in-domain data, NN-Rank beats state-of-the-art baselines that leverage lexical and linguistic features, with average improvements of up to 35.56 NDCG for POS and 18.14 NDCG for NER. As prior approaches can fall back to language-level features if target language data is not available, we show that NN-Rank remains competitive using only the Bible, an out-of-domain corpus available for a large number of languages. Ablations on the amount of unlabeled target data show that, for subsets consisting of as few as 25 examples, NN-Rank produces high-quality rankings which achieve 92.8% of the NDCG achieved using all available target data for ranking.